Latent Variable Models of Selectional Preference
نویسنده
چکیده
This paper describes the application of so-called topic models to selectional preference induction. Three models related to Latent Dirichlet Allocation, a proven method for modelling document-word cooccurrences, are presented and evaluated on datasets of human plausibility judgements. Compared to previously proposed techniques, these models perform very competitively, especially for infrequent predicate-argument combinations where they exceed the quality of Web-scale predictions while using relatively little data.
منابع مشابه
Improving Lexical Semantics for Sentential Semantics: Modeling Selectional Preference and Similar Words in a Latent Variable Model
Sentence Similarity [SS] computes a similarity score between two sentences. The SS task differs from document level semantics tasks in that it features the sparsity of words in a data unit, i.e. a sentence. Accordingly it is crucial to robustly model each word in a sentence to capture the complete semantic picture of the sentence. In this paper, we hypothesize that by better modeling lexical se...
متن کاملComputational Models for Chinese Selectional Preferences Induction
Selectional preference (SP) is an important kind of semantic knowledge. It can be used in various natural language processing tasks, including metaphor computing, lexicon building, syntactic structure disambiguation, word sense disambiguation, semantic role labeling, anaphora resolution, etc. This paper presents and compares two computational models for Chinese SP induction, a HowNet-based Sele...
متن کاملLearning Semantics and Selectional Preference of Adjective-Noun Pairs
We investigate the semantic relationship between a noun and its adjectival modifiers. We introduce a class of probabilistic models that enable us to to simultaneously capture both the semantic similarity of nouns and modifiers, and adjective-noun selectional preference. Through a combination of novel and existing evaluations we test the degree to which adjective-noun relationships can be catego...
متن کاملDomain Adaptation of a Dependency Parser with a Class-Class Selectional Preference Model
When porting parsers to a new domain, many of the errors are related to wrong attachment of out-of-vocabulary words. Since there is no available annotated data to learn the attachment preferences of the target domain words, we attack this problem using a model of selectional preferences based on domainspecific word classes. Our method uses Latent Dirichlet Allocations (LDA) to learn a domain-sp...
متن کاملProbabilistic Distributional Semantics with Latent Variable Models
We describe a probabilistic framework for acquiring selectional preferences of linguistic predicates and for using the acquired representations to model the effects of context on word meaning. Our framework uses Bayesian latent-variable models inspired by, and extending, the well-known Latent Dirichlet Allocation (LDA) model of topical structure in documents; when applied to predicate–argument ...
متن کامل